Hypothesis spaces for minimum Bayes risk training in large vocabulary speech recognition
نویسندگان
چکیده
The Minimum Bayes Risk (MBR) framework has been a successful strategy for the training of hidden Markov models for large vocabulary speech recognition. Practical implementations of MBR must select an appropriate hypothesis space and loss function. The set of word sequences and a word-based Levenshtein distance may be assumed to be the optimal choice but use of phoneme-based criteria appears to be more successful. This paper compares the use of different hypothesis spaces and loss functions defined using the system constituents of word, phone, physical triphone, physical state and physical mixture component. For practical reasons the competing hypotheses are constrained by sampling. The impact of the sampling technique on the performance of MBR training is also examined.
منابع مشابه
Hypothesis Spaces For Minimum Bayes R Speech Recogn
The Minimum Bayes Risk (MBR) framework has been a successful strategy for the training of hidden Markov models for large vocabulary speech recognition. Practical implementations of MBR must select an appropriate hypothesis space and loss function. The set of word sequences and a word-based Levenshtein distance may be assumed to be the optimal choice but use of phoneme-based criteria appears to ...
متن کاملLattice segmentation and minimum Bayes risk discriminative training for large vocabulary continuous speech recognition
Lattice segmentation techniques developed for Minimum Bayes Risk decoding in large vocabulary speech recognition tasks are used to compute the statistics needed for discriminative training algorithms that estimate HMM parameters so as to reduce the overall risk over the training data. New estimation procedures are developed and evaluated for both small and large vocabulary recognition tasks, an...
متن کاملBoosting Minimum Bayes Risk Discriminative Training
A new variant of AdaBoost is applied to a Minimum Bayes Risk discriminative training procedure that directly aims at reducing Word Error Rate for Automatic Speech Recognition. Both techniques try to improve the discriminative power of a classifier and we show that can be combined together to yield even better performance on a small vocabulary continuous speech recognition task. Our results also...
متن کاملGinisupport vector machines for segmental minimum Bayes risk decoding of continuous speech
We describe the use of Support Vector Machines (SVMs) for continuous speech recognition by incorporating them in Segmental Minimum Bayes Risk decoding. Lattice cutting is used to convert the Automatic Speech Recognition search space into sequences of smaller recognition problems. SVMs are then trained as discriminative models over each of these problems and used in a rescoring framework. We pos...
متن کاملPinched lattice minimum Bayes risk discriminative training for large vocabulary continuous speech recognition
Iterative estimation procedures that minimize empirical risk based on general loss functions such as the Levenshtein distance have been derived as extensions of the Extended Baum Welch algorithm. While reducing expected loss on training data is a desirable training criterion, these algorithms can be difficult to apply. They are unlike MMI estimation in that they require an explicit listing of t...
متن کامل